Goto

Collaborating Authors

 feedback law


Characterizing Human Feedback-Based Control in Naturalistic Driving Interactions via Gaussian Process Regression with Linear Feedback

DiPirro, Rachel, Devonport, Rosalyn, Calderone, Dan, Yang, Chishang "Mario'', Ju, Wendy, Oishi, Meeko

arXiv.org Artificial Intelligence

Understanding driver interactions is critical to designing autonomous vehicles to interoperate safely with human-driven cars. We consider the impact of these interactions on the policies drivers employ when navigating unsigned intersections in a driving simulator. The simulator allows the collection of naturalistic decision-making and behavior data in a controlled environment. Using these data, we model the human driver responses as state-based feedback controllers learned via Gaussian Process regression methods. We compute the feedback gain of the controller using a weighted combination of linear and nonlinear priors. We then analyze how the individual gains are reflected in driver behavior. We also assess differences in these controllers across populations of drivers. Our work in data-driven analyses of how drivers determine their policies can facilitate future work in the design of socially responsive autonomy for vehicles.


Integrator Forwading Design for Unicycles with Constant and Actuated Velocity in Polar Coordinates

Krstic, Miroslav, Todorovski, Velimir, Kim, Kwang Hak, Astolfi, Alessandro

arXiv.org Artificial Intelligence

Abstract-- In a companion paper, we present a modular framework for unicycle stabilization in polar coordinates that provides smooth steering laws through backstepping. Surprisingly, the same problem also allows application of integrator forwarding. In this work, we leverage this feature and construct new smooth steering laws together with control Lyapunov functions (CLFs), expanding the set of CLFs available for inverse optimal control design. In the case of constant forward velocity (Dubins car), backstepping produces finite-time (deadbeat) parking, and we show that integrator forwarding yields the very same class of solutions. This reveals a fundamental connection between backstepping and forwarding in addressing both the unicycle and, the Dubins car parking problems.


Modular Design of Strict Control Lyapunov Functions for Global Stabilization of the Unicycle in Polar Coordinates

Todorovski, Velimir, Kim, Kwang Hak, Krstic, Miroslav

arXiv.org Artificial Intelligence

Abstract-- Since the mid-1990s, it has been known that, unlike in Cartesian form where Brockett's condition rules out static feedback stabilization, the unicycle is globally asymptotically stabilizable by smooth feedback in polar coordinates. In this note, we introduce a modular framework for designing smooth feedback laws that achieve global asymptotic stabilization in polar coordinates. These laws are bidirectional, enabling efficient parking maneuvers, and are paired with families of strict control Lyapunov functions (CLFs) constructed in a modular fashion. The resulting CLFs guarantee global asymptotic stability with explicit convergence rates and include barrier variants that yield "almost global" stabilization, excluding only zero-measure subsets of the rotation manifolds. The strictness of the CLFs is further leveraged in our companion paper, where we develop inverse-optimal redesigns with meaningful cost functions and infinite gain margins.


A Long-Duration Autonomy Approach to Connected and Automated Vehicles

Beaver, Logan E.

arXiv.org Artificial Intelligence

A Long-Duration Autonomy Approach to Connected and Automated V ehicles Logan E. Beaver, Member, IEEE Abstract --In this article, we present a long-duration autonomy approach for the control of connected and automated vehicles (CA Vs) operating in a transportation network. In particular, we focus on the performance of CA Vs at traffic bottlenecks, including roundabouts, merging roadways, and intersections. We take a principled approach based on optimal control, and derive a reactive controller with guarantees on safety, performance, and energy efficiency. We guarantee safety through high order control barrier functions (HOCBFs), which we "lift" to first order CBFs using time-optimal motion primitives. We demonstrate the performance of our approach in simulation and compare it to an optimal control-based approach. Index T erms --autonomous systems, connected vehicles, long-duration autonomy, barrier functions I. I NTRODUCTION C ONNECTED and automated vehicles (CA Vs) continue to proliferate transportation networks. As a result, it is critical for us to develop control algorithms that are computationally efficient, provably safe, and produce energy-efficient trajectories.


Task-Space Riccati Feedback based Whole Body Control for Underactuated Legged Locomotion

Yang, Shunpeng, Hong, Zejun, Li, Sen, Wensing, Patrick, Zhang, Wei, Chen, Hua

arXiv.org Artificial Intelligence

This manuscript primarily aims to enhance the performance of whole-body controllers(WBC) for underactuated legged locomotion. We introduce a systematic parameter design mechanism for the floating-base feedback control within the WBC. The proposed approach involves utilizing the linearized model of unactuated dynamics to formulate a Linear Quadratic Regulator(LQR) and solving a Riccati gain while accounting for potential physical constraints through a second-order approximation of the log-barrier function. And then the user-tuned feedback gain for the floating base task is replaced by a new one constructed from the solved Riccati gain. Extensive simulations conducted in MuJoCo with a point bipedal robot, as well as real-world experiments performed on a quadruped robot, demonstrate the effectiveness of the proposed method. In the different bipedal locomotion tasks, compared with the user-tuned method, the proposed approach is at least 12% better and up to 50% better at linear velocity tracking, and at least 7% better and up to 47% better at angular velocity tracking. In the quadruped experiment, linear velocity tracking is improved by at least 3% and angular velocity tracking is improved by at least 23% using the proposed method.


Multi-Contact Whole Body Force Control for Position-Controlled Robots

Rouxel, Quentin, Ivaldi, Serena, Mouret, Jean-Baptiste

arXiv.org Artificial Intelligence

Many humanoid and multi-legged robots are controlled in positions rather than in torques, preventing direct control of contact forces, and hampering their ability to create multiple contacts to enhance their balance, such as placing a hand on a wall or a handrail. This paper introduces the SEIKO (Sequential Equilibrium Inverse Kinematic Optimization) pipeline, drawing inspiration from flexibility models used in serial elastic actuators to indirectly control contact forces on traditional position-controlled robots. SEIKO formulates whole-body retargeting from Cartesian commands and admittance control using two quadratic programs solved in real time. We validated our pipeline with experiments on the real, full-scale humanoid robot Talos in various multicontact scenarios, including pushing tasks, far-reaching tasks, stair climbing, and stepping on sloped surfaces. This work opens the possibility of stable, contact-rich behaviors while getting around many of the challenges of torque-controlled robots. Code and videos are available at https://hucebot.github.io/seiko\_controller\_website/.


Neighboring Extremal Optimal Control Theory for Parameter-Dependent Closed-loop Laws

Rai, Ayush, Mou, Shaoshuai, Anderson, Brian D. O.

arXiv.org Artificial Intelligence

This study introduces an approach to obtain a neighboring extremal optimal control (NEOC) solution for a closed-loop optimal control problem, applicable to a wide array of nonlinear systems and not necessarily quadratic performance indices. The approach involves investigating the variation incurred in the functional form of a known closed-loop optimal control law due to small, known parameter variations in the system equations or the performance index. The NEOC solution can formally be obtained by solving a linear partial differential equation, akin to those encountered in the iterative solution of a nonlinear Hamilton-Jacobi equation. Motivated by numerical procedures for solving these latter equations, we also propose a numerical algorithm based on the Galerkin algorithm, leveraging the use of basis functions to solve the underlying Hamilton-Jacobi equation of the original optimal control problem. The proposed approach simplifies the NEOC problem by reducing it to the solution of a simple set of linear equations, thereby eliminating the need for a full re-solution of the adjusted optimal control problem. Furthermore, the variation to the optimal performance index can be obtained as a function of both the system state and small changes in parameters, allowing the determination of the adjustment to an optimal control law given a small adjustment of parameters in the system or the performance index. Moreover, in order to handle large known parameter perturbations, we propose a homotopic approach that breaks down the single calculation of NEOC into a finite set of multiple steps. Finally, the validity of the claims and theory is supported by theoretical analysis and numerical simulations.


On the Feedback Law in Stochastic Optimal Nonlinear Control

Mohamed, Mohamed Naveed Gul, Chakravorty, Suman, Goyal, Raman, Wang, Ran

arXiv.org Artificial Intelligence

We consider the problem of nonlinear stochastic optimal control. This problem is thought to be fundamentally intractable owing to Bellman's infamous "curse of dimensionality". We present a result that shows that repeatedly solving an open-loop deterministic problem from the current state, similar to Model Predictive Control (MPC), results in a feedback policy that is $O(\epsilon^4)$ near to the true global stochastic optimal policy. Furthermore, empirical results show that solving the Stochastic Dynamic Programming (DP) problem is highly susceptible to noise, even when tractable, and in practice, the MPC-type feedback law offers superior performance even for stochastic systems.


Dimensionless Policies based on the Buckingham $\pi$ Theorem: Is it a good way to Generalize Numerical Results?

Girard, Alexandre

arXiv.org Artificial Intelligence

Yes if the context, the list of variables defining the motion control problem, is dimensionally similar. Here we show that by modifying the problem formulation using dimensionless variables, we can re-use the optimal control law generated numerically for a specific system to a sub-space of dimensionally similar systems. This is demonstrated, with numerically generated optimal controllers, for the classic motion control problem of swinging-up a torque-limited inverted pendulum. We also discuss the concept of regime, a region in the space of context variables, that can help relax the condition on dimensional similarity. Futhermore, we discuss how applying dimensionnal scaling of the input and output of a context-specific policy is equivalent to substituing the new systems parameters in an analytical equation for dimentionnaly similar systems. It remains to be seen if this approach can also help generalizing policies for more complex high-dimensional problems.


Transporting Robotic Swarms via Mean-Field Feedback Control

Zheng, Tongjia, Han, Qing, Lin, Hai

arXiv.org Artificial Intelligence

With the rapid development of AI and robotics, transporting a large swarm of networked robots has foreseeable applications in the near future. Existing research in swarm robotics has mainly followed a bottom-up philosophy with predefined local coordination and control rules. However, it is arduous to verify the global requirements and analyze their performance. This motivates us to pursue a top-down approach, and develop a provable control strategy for deploying a robotic swarm to achieve a desired global configuration. Specifically, we use mean-field partial differential equations (PDEs) to model the swarm and control its mean-field density (i.e., probability density) over a bounded spatial domain using mean-field feedback. The presented control law uses density estimates as feedback signals and generates corresponding velocity fields that, by acting locally on individual robots, guide their global distribution to a target profile. The design of the velocity field is therefore centralized, but the implementation of the controller can be fully distributed -- individual robots sense the velocity field and derive their own velocity control signals accordingly. The key contribution lies in applying the concept of input-to-state stability (ISS) to show that the perturbed closed-loop system (a nonlinear and time-varying PDE) is locally ISS with respect to density estimation errors. The effectiveness of the proposed control laws is verified using agent-based simulations.